{37 () Bayesian Network Classiiers. *

نویسندگان

  • DAN GEIGER
  • M. GOLDSZMIDT
چکیده

Recent work in supervised learning has shown that a surprisingly simple Bayesian classiier with strong assumptions of independence among features, called naive Bayes, is competitive with state-of-the-art classiiers such as C4.5. This fact raises the question of whether a classiier with less restrictive assumptions can perform even better. In this paper we evaluate approaches for inducing classiiers from data, based on the theory of learning Bayesian networks. These networks are factored representations of probability distributions that generalize the naive Bayesian classiier and explicitly represent statements about independence. Among these approaches we single out a method we call Tree Augmented Naive Bayes (TAN), which outperforms naive Bayes, yet at the same time maintains the computational simplicity (no search involved) and robustness that characterize naive Bayes. We experimentally tested these approaches, using problems from the University of California at Irvine repository, and compared them to C4.5, naive Bayes, and wrapper methods for feature selection.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Induction of Selective Bayesian Network Classiiers

We present an algorithm for inducing Bayesian networks using feature selection. The algorithm selects a subset of attributes that maximizes predictive accuracy prior to the network learning phase, thereby incorporating a bias for small networks that retain high predictive accuracy. We compare the behavior of this selective Bayesian network classiier with that of (a) Bayesian network classiiers ...

متن کامل

Bayesian Classifiers Are Large Margin Hyperplanes in a Hilbert Space

Bayesian algorithms for Neural Networks are known to produce classiiers which are very resistant to overrtting. It is often claimed that one of the main distinctive features of Bayesian Learning Algorithms is that they don't simply output one hypothesis, but rather an entire distribution of probability over an hypothesis set: the Bayes posterior. An alternative perspective is that they output a...

متن کامل

Biological Data Mining Using Bayesian Neural Networks: A Case Study

Biological data mining is the activity of nding signiicant information in biomolecular data. The signiicant information may refer to motifs, clusters, genes, and protein signatures. This paper presents an example of biological data mining: the recognition of promoters in DNA. We propose a two-level ensemble of classiiers to recognize E. Coli promoter sequences. The rst-level classiiers include ...

متن کامل

Recognizing Promoters in DNA Using Bayesian Neural Networks

Binary data classiication is to recognize positive data from unlabeled test data that may contain both positive and negative data. In this paper we propose a two-level approach to recognize E. Coli promoters in unlabeled DNA containing both promoter and non-promoter sequences. The rst-level classiiers include three Bayesian neural networks which learn from three diierent feature sets. The outpu...

متن کامل

Bayesian Voting Schemes as Large MarginClassi

It is often claimed that one of the main distinctive features of Bayesian Learning Algorithms for neural networks is that they don't simply output one hypothesis, but rather an entire distribution of probability over an hypothesis set: the Bayes posterior. An alternative perspective is that they output a linear combination of classiiers, whose coeecients are given by Bayes theorem. This can be ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1997